Scalability of Enhanced Parallel Batch Pattern BP Training Algorithm on General-Purpose Supercomputers
نویسندگان
چکیده
The development of an enhanced parallel algorithm for batch pattern training of a multilayer perceptron with the back propagation training algorithm and the research of its efficiency on general-purpose parallel computers are presented in this paper. An algorithmic description of the parallel version of the batch pattern training method is described. Several technical solutions which lead to enhancement of the parallelization efficiency of the algorithm are discussed. The efficiency of parallelization of the developed algorithm is investigated by progressively increasing the dimension of the parallelized problem on two generalpurpose parallel computers. The results of the experimental researches show that (i) the enhanced version of the parallel algorithm is scalable and provides better parallelization efficiency than the old implementation; (ii) the parallelization efficiency of the algorithm is high enough for an efficient use of this algorithm on general-purpose parallel computers available within modern computational grids.
منابع مشابه
Scalability of Parallel Batch Pattern Neural Network Training Algorithm
The development of parallel batch pattern back propagation training algorithm of multilayer perceptron and its scalability research on general-purpose parallel computer are presented in this paper. The model of multilayer perceptron and batch pattern training algorithm are theoretically described. The algorithmic description of the parallel batch pattern training method is presented. The scalab...
متن کاملImprovement of parallelization efficiency of batch pattern BP training algorithm using Open MPI
The use of tuned collective’s module of Open MPI to improve a parallelization efficiency of parallel batch pattern back propagation training algorithm of a multilayer perceptron is considered in this paper. The multilayer perceptron model and the usual sequential batch pattern training algorithm are theoretically described. An algorithmic description of a parallel version of the batch pattern t...
متن کاملAutomatic Self-Allocating Threads (ASAT) on the Convex Exemplar
Parallel processing systems have an advantage over traditional supercomputers in price/performance, but traditional supercomputers retain a signiicant advantage over parallel processing systems in the area of exibility. Traditional supercomputers can easily handle a mix of interactive, batch, scalar, vector, parallel, and large memory jobs simultaneously while maintaining high utilization. Ofte...
متن کاملShift-Based Parallel Image Compositing on InfiniBandTM Fat-Trees
Parallel image compositing has been widely studied over the past 20 years, as this is one, if not the most, crucial element in the implementation of a scalable parallel rendering system. Many algorithms have been proposed and implemented on a large variety of supercomputers. Among the existing supercomputers, InfiniBandTM (IB) PC clusters, and their associated fat-tree topology, are clearly bec...
متن کاملA New Parallel Matrix Multiplication Algorithm for Wormhole-Routed All-Port 2D/3D Torus Networks
A new matrix multiplication algorithm is proposed for massively parallel supercomputers with 2D/3D, all-port torus interconnection networks. The proposed algorithm is based on the traditional row-by-column multiplication matrix product model and employs a special routing pattern for better scalability. It compares favorably to the variants of Cannon’s and DNS algorithms since it allows matrices...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2010